- Home
- Search Results
- Page 1 of 1
Search for: All records
-
Total Resources5
- Resource Type
-
0002000003000000
- More
- Availability
-
23
- Author / Contributor
- Filter by Author / Creator
-
-
Booth, Joshua Dennis (4)
-
Sun, Hongyang (3)
-
Raghavan, Padma (2)
-
Bolet, Gregory S. (1)
-
Booth, Joshua (1)
-
Chen, Zizhao (1)
-
Garnett, Trevor (1)
-
Johnson, Daniel Ryley (1)
-
Lane, Phillip (1)
-
Verrecchia, Thomas (1)
-
#Tyler Phillips, Kenneth E. (0)
-
#Willis, Ciara (0)
-
& Abreu-Ramos, E. D. (0)
-
& Abramson, C. I. (0)
-
& Abreu-Ramos, E. D. (0)
-
& Adams, S.G. (0)
-
& Ahmed, K. (0)
-
& Ahmed, Khadija. (0)
-
& Aina, D.K. Jr. (0)
-
& Akcil-Okan, O. (0)
-
- Filter by Editor
-
-
& Spizer, S. M. (0)
-
& . Spizer, S. (0)
-
& Ahn, J. (0)
-
& Bateiha, S. (0)
-
& Bosch, N. (0)
-
& Brennan K. (0)
-
& Brennan, K. (0)
-
& Chen, B. (0)
-
& Chen, Bodong (0)
-
& Drown, S. (0)
-
& Ferretti, F. (0)
-
& Higgins, A. (0)
-
& J. Peters (0)
-
& Kali, Y. (0)
-
& Ruiz-Arias, P.M. (0)
-
& S. Spitzer (0)
-
& Sahin. I. (0)
-
& Spitzer, S. (0)
-
& Spitzer, S.M. (0)
-
(submitted - in Review for IEEE ICASSP-2024) (0)
-
-
Have feedback or suggestions for a way to improve these results?
!
Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
-
Parallelizing code in a shared-memory environment is commonly done utilizing loop scheduling (LS) in a fork-join manner as in OpenMP. This manner of parallelization is popular due to its ease to code, but the choice of the LS method is important when the workload per iteration is highly variable. Currently, the shared-memory environment is evolving in high-performance computing as larger chiplet-based processors with high core counts and segmented L3 cache are introduced. These processors have a stronger non-uniform memory access (NUMA) effect than the previous generation of x86-64 processors. This work attempts to modify the adaptive self-scheduling loop scheduler known asiCh(irregularChunk) for these NUMA environments while analyzing the impact of these systems on default OpenMP LS methods. In particular,iChis as a default LS method for irregular applications (i.e., applications where the workload per iteration is highly variable) that guarantees “good” performance without tuning. The modified version, namedNiCh, is demonstrated over multiple irregular applications to show the variation in performance. The work demonstrates thatNiChis able to better handle architectures with stronger NUMA effects, and particularly is better thaniChwhen the number of threads is greater than the number of cores. However,NiChalso comes with being less universally “good” thaniChand a set of parameters that are hardware dependent.more » « lessFree, publicly-accessible full text available December 31, 2025
-
Booth, Joshua Dennis; Sun, Hongyang; Garnett, Trevor (, Neural Computing and Applications)Free, publicly-accessible full text available January 1, 2026
-
Johnson, Daniel Ryley; Sun, Hongyang; Booth, Joshua Dennis; Raghavan, Padma (, IEEE)Free, publicly-accessible full text available November 13, 2025
-
Chen, Zizhao; Verrecchia, Thomas; Sun, Hongyang; Booth, Joshua; Raghavan, Padma (, ACM)
-
Booth, Joshua Dennis; Bolet, Gregory S. (, IEEE Access)
An official website of the United States government
